MobileLLM is an autoregressive language model optimized by Meta specifically for resource-constrained device-side applications. It adopts an optimized Transformer architecture and integrates key technologies such as the SwiGLU activation function, a deep and narrow architecture, embedding sharing, and grouped query attention. It performs excellently in zero-shot commonsense reasoning tasks.
Natural Language Processing
Transformers